Extending Local Learners with Error-correcting Output Codes Extending Local Learners with Error-correcting Output Codes
نویسندگان
چکیده
Error-correcting output codes (ECOCs) represent classes with a set of output bits, where each bit encodes a binary classiication task corresponding to a unique partition of the classes. Algorithms that use ECOCs learn the function corresponding to each bit, and combine them to generate class predictions. ECOCs can reduce both variance and bias errors for multiclass classiication tasks when the errors made at the output bits are not correlated. They work well with algorithms that eagerly induce global classiiers (e.g., C4.5) but do not assist simple local classiiers (e.g., nearest neighbor), which yield correlated predictions across the output bits. This is distressing because local learning algorithms for classiication are preferable to global classiiers for some types of applications. We show that the output bit predictions of local learners can be decorrelated by selecting diierent features for each bit. We present promising empirical results for this combination of ECOCs, nearest neighbor , and feature selection. We also describe modiications to the schemata racing algorithm for feature selection that improve its ability to retrieve good feature subsets in this context. Abstract. Error-correcting output codes (ECOCs) represent classes with a set of output bits, where each bit encodes a binary classiication task corresponding to a unique partition of the classes. Algorithms that use ECOCs learn the function corresponding to each bit, and combine them to generate class predictions. ECOCs can reduce both variance and bias errors for multiclass classiication tasks when the errors made at the output bits are not correlated. They work well with algorithms that eagerly induce global classiiers (e.g., C4.5) but do not assist simple local classiiers (e.g., nearest neighbor), which yield correlated predictions across the output bits. This is distressing because local learning algorithms for classiication are preferable to global classiiers for some types of applications. We show that the output bit predictions of local learners can be decorrelated by selecting diierent features for each bit. We present promising empirical results for this combination of ECOCs, nearest neighbor, and feature selection. We also describe modiications to the schemata racing algorithm for feature selection that improve its ability to retrieve good feature subsets in this context.
منابع مشابه
Bias , Variance , and Error Correcting Output Codes forLocal Learners ?
This paper focuses on a bias variance decomposition analysis of a local learning algorithm, the nearest neighbor classiier, that has been extended with error correcting output codes. This extended algorithm often considerably reduces the 0-1 (i.e., classiication) error in comparison with nearest neighbor (Ricci & Aha, 1997). The analysis presented here reveals that this performance improvement ...
متن کاملError-Correcting Output Codes for Local Learners
Error-correcting output codes (ECOCs) represent classes with a set of output bits, where each bit encodes a binary classiication task corresponding to a unique partition of the classes. Algorithms that use ECOCs learn the function corresponding to each bit, and combine them to generate class predictions. ECOCs can reduce both variance and bias errors for multiclass classiication tasks when the ...
متن کاملAn approach to fault detection and correction in design of systems using of Turbo codes
We present an approach to design of fault tolerant computing systems. In this paper, a technique is employed that enable the combination of several codes, in order to obtain flexibility in the design of error correcting codes. Code combining techniques are very effective, which one of these codes are turbo codes. The Algorithm-based fault tolerance techniques that to detect errors rely on the c...
متن کاملEnsemble Methods
The idea of ensemble learning is to employ multiple learners and combine their predictions. There is no definitive taxonomy. Jain, Duin and Mao (2000) list eighteen classifier combination schemes; Witten and Frank (2000) detail four methods of combining multiple models: bagging, boosting, stacking and errorcorrecting output codes whilst Alpaydin (2004) covers seven methods of combining multiple...
متن کاملOne-point Goppa Codes on Some Genus 3 Curves with Applications in Quantum Error-Correcting Codes
We investigate one-point algebraic geometric codes CL(D, G) associated to maximal curves recently characterized by Tafazolian and Torres given by the affine equation yl = f(x), where f(x) is a separable polynomial of degree r relatively prime to l. We mainly focus on the curve y4 = x3 +x and Picard curves given by the equations y3 = x4-x and y3 = x4 -1. As a result, we obtain exact value of min...
متن کامل